This chapter provides information about the software and services installed on Oracle Big Data Appliance. It contains these sections:
Monitoring Multiple Clusters Using Oracle Enterprise Manager
Collecting Diagnostic Information for Oracle Customer Support
An Oracle Enterprise Manager plug-in enables you to use the same system monitoring tool for Oracle Big Data Appliance as you use for Oracle Exadata Database Machine or any other Oracle Database installation. With the plug-in, you can view the status of the installed software components in tabular or graphic presentations, and start and stop these software services. You can also monitor the health of the network and the rack components.
Oracle Enterprise Manager enables you to monitor all Oracle Big Data Appliance racks on the same InfiniBand fabric. It provides summary views of both the rack hardware and the software layout of the logical clusters.
After opening Oracle Enterprise Manager web interface, logging in, and selecting a target cluster, you can drill down into these primary areas:
InfiniBand network: Network topology and status for InfiniBand switches and ports. See Figure 2-1.
Hadoop cluster: Software services for HDFS, MapReduce, and ZooKeeper.
Oracle Big Data Appliance rack: Hardware status including server hosts, Oracle Integrated Lights Out Manager (Oracle ILOM) servers, power distribution units (PDUs), and the Ethernet switch.
Figure 2-1 shows a small section of the cluster home page.
Figure 2-1 YARN Page in Oracle Enterprise Manager

To monitor Oracle Big Data Appliance using Oracle Enterprise Manager:
Download and install the plug-in. See Oracle Enterprise Manager System Monitoring Plug-in Installation Guide for Oracle Big Data Appliance.
Log in to Oracle Enterprise Manager as a privileged user.
From the Targets menu, choose Big Data Appliance to view the Big Data page. You can see the overall status of the targets already discovered by Oracle Enterprise Manager.
Select a target cluster to view its detail pages.
Expand the target navigation tree to display the components. Information is available at all levels.
Select a component in the tree to display its home page.
To change the display, choose an item from the drop-down menu at the top left of the main display area.
See Also:
Oracle Enterprise Manager System Monitoring Plug-in Installation Guide for Oracle Big Data Appliance for installation instructions and use cases.The Enterprise Manager command-line interface (emcli) is installed on Oracle Big Data Appliance along with all the other software. It provides the same functionality as the web interface. You must provide credentials to connect to Oracle Management Server.
To get help, enter emcli help.
Cloudera Manager is installed on Oracle Big Data Appliance to help you with Cloudera's Distribution including Apache Hadoop (CDH) operations. Cloudera Manager provides a single administrative interface to all Oracle Big Data Appliance servers configured as part of the Hadoop cluster.
Cloudera Manager simplifies the performance of these administrative tasks:
Monitor jobs and services
Start and stop services
Manage security and Kerberos credentials
Monitor user activity
Monitor the health of the system
Monitor performance metrics
Track hardware use (disk, CPU, and RAM)
Cloudera Manager runs on the ResourceManager node (node03) and is available on port 7180.
To use Cloudera Manager:
Open a browser and enter a URL like the following:
http://bda1node03.example.com:7180
In this example, bda1 is the name of the appliance, node03 is the name of the server, example.com is the domain, and 7180 is the default port number for Cloudera Manager.
Log in with a user name and password for Cloudera Manager. Only a user with administrative privileges can change the settings. Other Cloudera Manager users can view the status of Oracle Big Data Appliance.
See Also:
Cloudera Manager Monitoring and Diagnostics Guide ator click Help on the Cloudera Manager Support menu
In Cloudera Manager, you can choose any of the following pages from the menu bar across the top of the display:
Home: Provides a graphic overview of activities and links to all services controlled by Cloudera Manager. See Figure 2-2.
Clusters: Accesses the services on multiple clusters.
Hosts: Monitors the health, disk usage, load, physical memory, swap space, and other statistics for all servers in the cluster.
Diagnostics: Accesses events and logs. Cloudera Manager collects historical information about the systems and services. You can search for a particular phrase for a selected server, service, and time period. You can also select the minimum severity level of the logged messages included in the search: TRACE, DEBUG, INFO, WARN, ERROR, or FATAL.
Audits: Displays the audit history log for a selected time range. You can filter the results by user name, service, or other criteria, and download the log as a CSV file.
Charts: Enables you to view metrics from the Cloudera Manager time-series data store in a variety of chart types, such as line and bar.
Backup: Accesses snapshot policies and scheduled replications.
Administration: Provides a variety of administrative options, including Settings, Alerts, Users, and Kerberos.
Figure 2-2 shows the Cloudera Manager home page.
As a Cloudera Manager administrator, you can change various properties for monitoring the health and use of Oracle Big Data Appliance, add users, and set up Kerberos security.
To access Cloudera Manager Administration:
Log in to Cloudera Manager with administrative privileges.
Select the Administration tab.
Cloudera Manager provides the interface for managing these services:
HDFS
Hive
Hue
Oozie
YARN
ZooKeeper
You can use Cloudera Manager to change the configuration of these services, stop, and restart them. Additional services are also available, which require configuration before you can use them. See "Unconfigured Software."
Note:
Manual edits to Linux service scripts or Hadoop configuration files do not affect these services. You must manage and configure them using Cloudera Manager.You also have the option of using the native Hadoop utilities. These utilities are read-only and do not require authentication.
Cloudera Manager provides an easy way to obtain the correct URLs for these utilities. On the YARN service page, expand the Web UI submenu.
You can monitor MapReduce jobs using the resource manager interface.
To monitor MapReduce jobs:
Open a browser and enter a URL like the following:
http://bda1node03.example.com:8088
In this example, bda1 is the name of the rack, node03 is the name of the server where the YARN resource manager runs, and 8088 is the default port number for the user interface.
Figure 2-3 shows the resource manager interface.
Figure 2-3 YARN Resource Manager Interface

You can monitor the health of the Hadoop file system by using the DFS health utility on the first two nodes of a cluster.
To monitor HDFS:
Open a browser and enter a URL like the following:
http://bda1node01.example.com:50070
In this example, bda1 is the name of the rack, node01 is the name of the server where the dfshealth utility runs, and 50070 is the default port number for the user interface.
Figure 2-3 shows the DFS health utility interface.
Hue runs in a browser and provides an easy-to-use interface to several applications to support interaction with Hadoop and HDFS. You can use Hue to perform any of the following tasks:
Query Hive data stores
Create, load, and delete Hive tables
Work with HDFS files and directories
Create, submit, and monitor MapReduce jobs
Monitor MapReduce jobs
Create, edit, and submit workflows using the Oozie dashboard
Manage users and groups
Hue is automatically installed and configured on Oracle Big Data Appliance. It runs on port 8888 of the ResourceManager node (node03).
To use Hue:
Log in to Cloudera Manager and click the hue service on the Home page.
On the hue page, click Hue Web UI.
Bookmark the Hue URL, so that you can open Hue directly in your browser. The following URL is an example:
http://bda1node03.example.com:8888
Log in with your Hue credentials.
Oracle Big Data Appliance is not configured initially with any Hue user accounts. The first user who connects to Hue can log in with any user name and password, and automatically becomes an administrator. This user can create other user and administrator accounts.
Figure 2-5 shows the Hive Query Editor.
The following sections identify the software installed on Oracle Big Data Appliance. Some components operate with Oracle Database 11.2.0.2 and later releases.
This section contains the following topics:
These software components are installed on all servers in the cluster. Oracle Linux, required drivers, firmware, and hardware verification utilities are factory installed. All other software is installed on site. The optional software components may not be configured in your installation.
Note:
You do not need to install additional software on Oracle Big Data Appliance. Doing so may result in a loss of warranty and support. See the Oracle Big Data Appliance Owner's Guide.Base image software:
Oracle Linux 6.4 (upgrades stay at 5.8) with Oracle Unbreakable Enterprise Kernel version 2 (UEK2)
Java HotSpot Virtual Machine 7 version 25 (JDK 7u25)
Oracle R Distribution 3.0.1-2
MySQL Database 5.5.35 Advanced Edition
Puppet, firmware, Oracle Big Data Appliance utilities
Oracle InfiniBand software
Mammoth installation:
Cloudera's Distribution including Apache Hadoop Release 5 (5.0) including:
Apache Hive 0.12
Apache HBase
Apache Spark
Cloudera Search 1.2.0
Oracle NoSQL Database Community Edition or Enterprise Edition 12c Release 1 Version 3.0.5 (optional)
Oracle Big Data Connectors 3.0 (optional):
See Also:
Oracle Big Data Appliance Owner's Guide for information about the Mammoth utilityFigure 2-6 shows the relationships among the major components.
Figure 2-6 Major Software Components of Oracle Big Data Appliance

Your Oracle Big Data Appliance license includes all components in Cloudera Enterprise Data Hub Edition. All CDH components are installed automatically by the Mammoth utility. Do not download them from the Cloudera website.
See Also:
For a complete list of CDH components:http://www.cloudera.com/content/cloudera/en/products-and-services/product-comparison.html
However, you must use Cloudera Manager to add the following services before you can use them:
Cloudera Search (Solr)
To add a service:
Log in to Cloudera Manager as the admin user.
On the Home page, expand the cluster menu in the left panel and choose Add a Service to open the Add Service wizard. The first page lists the services you can add.
Follow the steps of the wizard.
You can find the RPM files on the first server of each cluster in /opt/oracle/BDAMammoth/bdarepo/RPMS/noarch.
See Also:
CDH5 Installation and Configuration Guide for configuration procedures atAll services are installed on all nodes in a CDH cluster, but individual services run only on designated nodes. There are slight variations in the location of the services depending on the configuration of the cluster.
This section describes the services in a default YARN configuration.
This section contains the following topics:
Table 2-1 identifies the services in CDH clusters configured within a single rack, including starter racks and clusters with six nodes. Node01 is the first server in the cluster (server 1, 7, or 10), and nodenn is the last server in the cluster (server 6, 9, 12, or 18).
Table 2-1 Service Locations for One or More CDH Clusters in a Single Rack
| Node01 | Node02 | Node03 | Node04 | Node05 to nn |
|---|---|---|---|---|
|
Balancer |
Cloudera Manager Server |
|||
|
Cloudera Manager Agent |
Cloudera Manager Agent |
Cloudera Manager Agent |
Cloudera Manager Agent |
Cloudera Manager Agent |
|
DataNode |
DataNode |
DataNode |
DataNode |
DataNode |
|
Failover Controller |
Failover Controller |
JobHistory |
Hive, Hue, Oozie, Solr |
|
|
JournalNode |
JournalNode |
JournalNode |
||
|
MySQL Backup |
MySQL Primary |
|||
|
NameNode |
NameNode |
|||
|
NodeManagerFoot 1 |
NodeManagerFootref 1 |
NodeManager |
NodeManager |
NodeManager |
|
Oracle Data Integrator Agent |
||||
|
Puppet |
Puppet |
Puppet |
Puppet |
Puppet |
|
Puppet Master |
ResourceManager |
ResourceManager |
||
|
ZooKeeper |
ZooKeeper |
ZooKeeper |
Footnote 1 Starter racks and six-node clusters only, with reduced allocated resources
When multiple racks are configured as a single CDH cluster, some critical services are installed on the second rack.
Table 2-2 identifies the location of services on the first rack of a multirack cluster.
Table 2-2 Service Locations in the First Rack of a Multirack CDH Cluster
| Node01 | Node02 | Node03 | Node04 | Node05 to nnFoot 1 |
|---|---|---|---|---|
|
Cloudera Manager Agent |
Cloudera Manager Agent |
Cloudera Manager Agent |
Cloudera Manager Agent |
Cloudera Manager Agent |
|
Cloudera Manager Server |
||||
|
DataNode |
DataNode |
DataNode |
DataNode |
DataNode |
|
Failover Controller |
||||
|
JournalNode |
JournalNode |
|||
|
NameNode |
MySQL Primary |
|||
|
NodeManager |
NodeManager |
NodeManager |
NodeManager |
NodeManager |
|
Puppet |
Puppet |
Puppet |
Puppet |
Puppet |
|
Puppet Master |
ResourceManager |
|||
|
ZooKeeper |
ZooKeeper |
Footnote 1 nn includes the servers in additional racks.
Table 2-3 shows the service locations in the second rack of a multirack cluster.
Table 2-3 Service Locations in the Second Rack of a Multirack CDH Cluster
| Node01 | Node02 | Node03 | Node04 | Node05 to nn |
|---|---|---|---|---|
|
Balancer |
||||
|
Cloudera Manager Agent |
Cloudera Manager Agent |
Cloudera Manager Agent |
Cloudera Manager Agent |
Cloudera Manager Agent |
|
DataNode |
DataNode |
DataNode |
DataNode |
DataNode |
|
Failover Controller |
||||
|
JournalNode |
Hive, Hue, Oozie, Solr |
|||
|
MySQL Backup |
||||
|
NameNode |
||||
|
NodeManagerFoot 1 |
NodeManagerFootref 1 |
NodeManager |
NodeManager |
NodeManager |
|
Oracle Data Integrator Agent |
||||
|
Puppet |
Puppet |
Puppet |
Puppet |
Puppet |
|
Puppet Master |
ResourceManager |
|||
|
ZooKeeper |
Footnote 1 Starter racks and six-node clusters only, with reduced allocated resources
Yet Another Resource Negotiator (YARN) is the default version of MapReduce, starting with Oracle Big Data Appliance 3.0. Oracle recommends YARN for production systems. MapReduce 1 (MRv1) is supported for backward compatibility, but you must install it manually. Do not run both versions of MapReduce (YARN and MRv1) on the same cluster.
MapReduce applications developed using MapReduce 1 may require recompilation to run under YARN.
The ResourceManager performs all resource management tasks. An MRAppMaster performs the job management tasks. Each job has its own MRAppMaster. The NodeManager has containers that can run a map task, a reduce task, or an MRAppMaster. The NodeManager can dynamically allocate containers using the available memory. This architecture results in improved scalability and better use of the cluster than MRv1.
YARN also manages resources for Spark and Impala.
MRv1 manages MapReduce jobs exclusively. The JobTracker combines the functions of the ResourceManager and MRAppMaster. TaskTrackers perform the function of the NodeManagers.
See Also:
"Running Existing Applications on Hadoop 2 YARN" athttp://hortonworks.com/blog/running-existing-applications-on-hadoop-2-yarn/
The NameNode is the most critical process because it keeps track of the location of all data. Without a healthy NameNode, the entire cluster fails. Apache Hadoop v0.20.2 and earlier are vulnerable to failure because they have a single name node.
Cloudera's Distribution including Apache Hadoop Version 4 (CDH5) reduces this vulnerability by maintaining redundant NameNodes. The data is replicated during normal operation as follows:
CDH maintains redundant NameNodes on the first two nodes of a cluster. One of the NameNodes is in active mode, and the other NameNode is in hot standby mode. If the active NameNode fails, then the role of active NameNode automatically fails over to the standby NameNode.
The NameNode data is written to a mirrored partition so that the loss of a single disk can be tolerated. This mirroring is done at the factory as part of the operating system installation.
The active NameNode records all changes to the file system metadata in at least two JournalNode processes, which the standby NameNode reads. There are three JournalNodes, which run on the first three nodes of each cluster.
The changes recorded in the journals are periodically consolidated into a single fsimage file in a process called checkpointing.
Note:
Oracle Big Data Appliance 2.0 and later releases do not support the use of an external NFS filer for backups and do not use NameNode federation.Figure 2-7 shows the relationships among the processes that support automatic failover of the NameNode.
Figure 2-7 Automatic Failover of the NameNode on Oracle Big Data Appliance

The ResourceManager allocates resources for application tasks and application masters across the cluster. Like the NameNode, the ResourceManager is a critical point of failure for the cluster. If all ResourceManagers fail, then all jobs stop running. Oracle Big Data Appliance 3.0 supports ResourceManager High Availability in Cloudera 5 to reduce this vulnerability.
CDH maintains redundant ResourceManager services on node03 and node04. One of the services is in active mode, and the other service is in hot standby mode. If the active service fails, then the role of active ResourceManager automatically fails over to the standby service. No failover controllers are required.
Figure 2-7 shows the relationships among the processes that support automatic failover of the ResourceManager.
Figure 2-8 Automatic Failover of the ResourceManager on Oracle Big Data Appliance

The NodeManager services are allocated a fixed amount of memory and CPU core voltage (Vcore), as shown in Table 2-4.
Table 2-4 Resource Allocation for NodeManagers
| Appliance Model | Memory | VCore |
|---|---|---|
|
Oracle Big Data Appliance X4-2 |
40 GB |
32V |
|
Oracle Big Data Appliance X3-2 |
40 GB |
32V |
|
Oracle Big Data Appliance (Sun Fire X4270 M2-based rack) |
30 GB |
24V |
Each node in a cluster has a maximum number of map and reduce tasks that are allowed to run simultaneously. Table 2-5 shows the default configuration of resources for the MapReduce service on Oracle Big Data Appliance X4-2 and X3-2.
Table 2-5 Maximum Map and Reduce Tasks on Oracle Big Data Appliance X3-2
| Node | 6-Node Cluster | Larger ClustersFoot 1 |
|---|---|---|
|
Node01 and Node02 |
14 map 8 reduce |
None (no NodeManager) |
|
Node03 and Node04 |
10 map 6 reduce |
10 map 6 reduce |
|
Node05 to Nodenn |
20 map 13 reduce |
20 map 13 reduce |
Footnote 1 9 or more nodes
Table 2-6 shows the default configuration of resources for the MapReduce service on Oracle Big Data Appliance Sun Fire X4270 M2-based racks.
Table 2-6 Maximum Map and Reduce Tasks on Sun Fire X4270 M2-Based Racks
| Node | 6-Node Cluster | Larger ClustersFoot 1 |
|---|---|---|
|
Node01 and Node02 |
10 map 6 reduce |
None |
|
Node03 and Node04 |
7 map 4 reduce |
7 map 4 reduce |
|
Node05 to Nodenn |
15 map 10 reduce |
15 map 10 reduce |
Footnote 1 9 or more nodes
The effects of a server failure vary depending on the server's function within the CDH cluster. Oracle Big Data Appliance servers are more robust than commodity hardware, so you should experience fewer hardware failures. This section highlights the most important services that run on the various servers of the primary rack. For a full list, see "Where Do the Services Run on a Single Rack?."
Note:
In a multirack cluster, some critical services run on the first server of the second rack. See "Where Do the Services Run on a Multirack Cluster?."Each server has 12 disks. The critical operating system is stored on disks 1 and 2.
Table 2-7 describes how the disks are partitioned.
Critical nodes are required for the cluster to operate normally and provide all services to users. In contrast, the cluster continues to operate with no loss of service when a noncritical node fails.
On single-rack clusters, the critical services are installed initially on the first four nodes of the cluster. The remaining nodes (node05 up to node18) only run noncritical services. If a hardware failure occurs on one of the critical nodes, then the services can be moved to another, noncritical server. For example, if node02 fails, then you might move its critical services node05. Table 2-1 identifies the initial location of services for clusters that are configured on a single rack.
In a multirack cluster, some critical services run on the first server of the second rack. See "Where Do the Services Run on a Single Rack?."
Moving a critical node requires that all clients be reconfigured with the address of the new node. The other alternative is to wait for the failed server to be repaired. You must weigh the loss of services against the inconvenience of reconfiguring the clients.
Table 2-8 Critical Service Locations
| Node Name | Initial Node Position | Critical Functions |
|---|---|---|
|
First NameNode |
Node01 |
ZooKeeperFoot 1 , first NameNodeFootref 1, NameNode failover controllerFootref 1, balancerFootref 1, puppet master |
|
Second NameNode |
Node02 |
ZooKeeper, second NameNode, NameNode failover controller, MySQL backup server |
|
First ResourceManager Node |
Node03 |
ZooKeeper, first ResourceManager, Cloudera Manager server, MySQL primary server |
|
Second ResourceManager Node |
Node04 |
Second ResourceManager, Oracle Data Integrator agent, Hue, Hive, Oozie |
Footnote 1 In multirack clusters, this service is initially installed on the first server of the second rack.
One instance of the NameNode initially runs on node01. If this node fails or goes offline (such as a restart), then the second NameNode (node02) automatically takes over to maintain the normal activities of the cluster.
Alternatively, if the second NameNode is already active, it continues without a backup. With only one NameNode, the cluster is vulnerable to failure. The cluster has lost the redundancy needed for automatic failover.
The puppet master also runs on this node. The Mammoth utility uses Puppet, and so you cannot install or reinstall the software if, for example, you must replace a disk drive elsewhere in the rack.
Note:
In multirack clusters, the NameNode service is installed on the first server of the second rack.One instance of the NameNode initially runs on node02. If this node fails, then the function of the NameNode either fails over to the first NameNode (node01) or continues there without a backup. However, the cluster has lost the redundancy needed for automatic failover if the first NameNode also fails.
The MySQL backup database also runs on this node. MySQL Database continues to run, although there is no backup of the master database.
One instance of the ResourceManager initially runs on node03. If this node fails or goes offline (such as a restart, then the second ResourceManager (node04) automatically takes over to distribute MapReduce tasks to specific nodes across the cluster.
Alternatively, if the second ResourceManager is already active, it continues without a backup. With only one ResourceManager, the cluster is vulnerable to failure. The cluster has lost the redundancy needed for automatic failover.
These services are also disrupted:
Cloudera Manager: This tool provides central management for the entire CDH cluster. Without this tool, you can still monitor activities using the utilities described in "Using Hadoop Monitoring Utilities".
MySQL Master Database: Cloudera Manager, Oracle Data Integrator, Hive, and Oozie use MySQL Database. The data is replicated automatically, but you cannot access it when the master database server is down.
One instance of the ResourceManager initially runs on node04. If this node fails, then the function of the ResourceManager either fails over to the first ResourceManager (node03) or continues there without a backup. However, the cluster has lost the redundancy needed for automatic failover if the first ResourceManager also fails.
These services are also disrupted:
Oracle Data Integrator: This service supports Oracle Data Integrator Application Adapter for Hadoop. You cannot use this connector when the ResourceManager node is down.
Hive: Hive provides a SQL-like interface to data that is stored in HDFS. Most of the Oracle Big Data Connectors can access Hive tables, which are not available if this node fails.
Hue: This administrative tool is not available when the ResourceManager node is down.
Oozie: This workflow and coordination service runs on the ResourceManager node, and is unavailable when the node is down.
The noncritical nodes (node05 to node18) are optional in that Oracle Big Data Appliance continues to operate with no loss of service if a failure occurs. The NameNode automatically replicates the lost data to always maintain three copies. MapReduce jobs execute on copies of the data stored elsewhere in the cluster. The only loss is in computational power, because there are fewer servers on which to distribute the work.
This section describes how to shut down Oracle Big Data Appliance gracefully and restart it.
You must have root access. Passwordless SSH must be set up on the cluster, so that you can use the dcli utility.
To ensure that passwordless-ssh is set up:
Log in to the first node of the cluster as root.
Use a dcli command to verify it is working. This command should return the IP address and host name of every node in the cluster:
# dcli -C hostname
192.0.2.1: bda1node01.example.com
192.0.2.2: bda1node02.example.com
.
.
.
If you do not get these results, then set up dcli on the cluster:
# setup-root-ssh -C
See Also:
Oracle Big Data Appliance Owner's Guide for details about these commands.Follow these procedures to shut down all Oracle Big Data Appliance software and hardware components.
Note:
The following services stop automatically when the system shuts down. You do not need to take any action:Oracle Enterprise Manager agent
Auto Service Request agents
Use Cloudera Manager to stop the services it manages, including flume, hbase, hdfs, hive, hue, mapreduce, oozie, and zookeeper.
Log in to Cloudera Manager as the admin user.
In the Status pane of the opening page, expand the menu for the cluster and click Stop, and then click Stop again when prompted to confirm. See Figure 2-9.
To navigate to this page, click the Home tab, and then the Status subtab.
On the Command Details page, click Close when all processes are stopped.
In the same pane under Cloudera Management Services, expand the menu for the mgmt service and click Stop.
Log out of Cloudera Manager.
Follow this procedure to stop Cloudera Manager Server.
Log in as root to the node where Cloudera Manager runs (initially node03).
Note:
The remaining tasks presume that you are logged in to a server asroot. You can enter the commands from any server by using the dcli command. This example runs the pwd command on node03 from any node in the cluster:
# dcli -c node03 pwd
Stop the Cloudera Manager server:
# service cloudera-scm-server stop Stopping cloudera-scm-server: [ OK ]
Verify that the server is stopped:
# service cloudera-scm-server status
cloudera-scm-server is stopped
After stopping Cloudera Manager, you cannot access it using the web console.
If Oracle Data Integrator Application Adapter for Hadoop is installed on the cluster, then stop the agent.
Check the status of the Oracle Data Integrator service:
# dcli -C service odi-agent status
Stop the Oracle Data Integrator agent, if it is running:
# dcli -C service odi-agent stop
Ensure that the Oracle Data Integrator service stopped running:
# dcli -C service odi-agent status
All nodes share an NFS directory on node03, and additional directories may also exist. If a server with the NFS directory (/opt/exportdir) is unavailable, then the other servers hang when attempting to shut down. Thus, you must dismount the NFS directories first.
Locate any mounted NFS directories:
# dcli -C mount | grep shareddir
192.0.2.1: bda1node03.example.com:/opt/exportdir on /opt/shareddir type nfs (rw,tcp,soft,intr,timeo=10,retrans=10,addr=192.0.2.3)
192.0.2.2: bda1node03.example.com:/opt/exportdir on /opt/shareddir type nfs (rw,tcp,soft,intr,timeo=10,retrans=10,addr=192.0.2.3)
192.0.2.3: /opt/exportdir on /opt/shareddir type none (rw,bind)
.
.
.
The sample output shows a shared directory on node03 (192.0.2.3).
Dismount the shared directory:
# dcli -C umount /opt/shareddir
Dismount any custom NFS directories.
The Linux shutdown -h command powers down individual servers. You can use the dcli -g command to stop multiple servers.
Create a file that lists the names or IP addresses of the other servers in the cluster, that is, not including the one you are logged in to.
Stop the other servers:
# dcli -g filename shutdown -h now
For filename, enter the name of the file that you created in step 1.
Stop the server you are logged in to:
# shutdown -h now
To stop the network switches, turn off a PDU or a breaker in the data center. The switches only turn off when power is removed.
The network switches do not have power buttons. They shut down only when power is removed
To stop the switches, turn off all breakers in the two PDUs.
Follow these procedures to power up the hardware and start all services on Oracle Big Data Appliance.
Switch on all 12 breakers on both PDUs.
Allow 4 to 5 minutes for Oracle ILOM and the Linux operating system to start on the servers.
If password-based, on-disk encryption is enabled, then log in and mount the Hadoop directories on those servers:
$ mount-hadoop-dirs Enter password to mount Hadoop directories: password
If the servers do not start automatically, then you can start them locally by pressing the power button on the front of the servers, or remotely by using Oracle ILOM. Oracle ILOM has several interfaces, including a command-line interface (CLI) and a web console. Use whichever interface you prefer.
For example, you can log in to the web interface as root and start the server from the Remote Power Control page. The URL for Oracle ILOM is the same as for the host, except that it typically has a -c or -ilom extension. This URL connects to Oracle ILOM for bda1node4:
http://bda1node04-ilom.example.com
Use Cloudera Manager to start all the HDFS services that it controls.
Log in as root to the node where Cloudera Manager runs (initially node03).
Note:
The remaining tasks presume that you are logged in to a server asroot. You can enter the commands from any server by using the dcli command. This example runs the pwd command on node03 from any node in the cluster:
# dcli -c node03 pwd
Verify that the Cloudera Manager started automatically on node03:
# service cloudera-scm-server status cloudera-scm-server (pid 11399) is running...
If it is not running, then start it:
# service cloudera-scm-server start
Log in to Cloudera Manager as the admin user.
In the Status pane of the opening page, expand the menu for the cluster and click Start, and then click Start again when prompted to confirm. See Figure 2-9.
To navigate to this page, click the Home tab, and then the Status subtab.
On the Command Details page, click Close when all processes are started.
In the same pane under Cloudera Management Services, expand the menu for the mgmt service and click Start.
Log out of Cloudera Manager (optional).
If Oracle Data Integrator Application Adapter for Hadoop is used on this cluster, then start the agent.
Check the status of the agent:
# /opt/oracle/odiagent/agent_standalone/oracledi/agent/bin/startcmd.sh OdiPingAgent [-AGENT_NAME=agent_name]
Start the agent:
# /opt/oracle/odiagent/agent_standalone/oracledi/agent/bin/agent.sh [-NAME=agent_name] [-PORT=port_number]
Oracle Big Data Appliance uses the Yet Another Resource Negotiator (YARN) implementation of MapReduce by default. You have the option of using classic MapReduce (MRv1) instead. You cannot use both implementations in the same cluster; you can activate either the MapReduce or the YARN service.
To switch a cluster to MRv1:
Log in to Cloudera Manager as the admin user.
Stop the YARN service:
Locate YARN in the list of services on the Status tab of the Home page.
Expand the YARN menu and click Stop.
On the cluster menu, click Add a Service to start the Add Service wizard:
Select MapReduce for the type of service you want to add.
Select hdfs/zookeeper as a dependency (default).
Customize the role assignments:
JobTracker: Click the field to display a list of nodes in the cluster. Select the third node.
TaskTracker: For a six-node cluster, keep the TaskTrackers on all nodes (default). For larger clusters, remove the TaskTrackers from the first two nodes.
On the Review Changes page, change the parameter values:
TaskTracker Local Data Directory List: Change the default group and group 1 to "/u12/hadoop/mapred".."/u01/hadoop/mapred".
JobTracker Local Data Directory List: Change the default group to "/u12/hadoop/mapred".."/u01/hadoop/mapred".
Complete the steps of the wizard with no further changes. Click Finish to save the configuration and exit.
Update the Hive service configuration:
On the Status tab of the Home page, click hive to display the hive page.
Expand the Configuration submenu and click View and Edit.
Select mapreduce as the value of the MapReduce Service property.
Click Save Changes.
Repeat step 4 to update the Oozie service configuration to use mapreduce.
On the Status tab of the Home page, expand the hive and oozie menus and choose Restart.
Optional: Expand the yarn service menu and choose Delete.
If you retain the yarn service, then after every cluster restart, you will see "Memory overcommit validation" warnings, and you must manually stop the yarn service.
Update the MapReduce service configuration:
On the Status tab of the Home page, click mapreduce to display the mapreduce page.
Expand the Configuration submenu and click View and Edit.
Under Category, expand TaskTracker Default Group, and then click Resource Management.
Set the following properties:
Java Heap Size of TaskTracker in Bytes: Reset to the default value of 1 GiB.
Maximum Number of Simultaneous Map Tasks: Set to either 15 for Sun Fire X4270 M2 racks or 20 for all other racks.
Maximum Number of Simultaneous Reduce Tasks: Set to either 10 for Sun Fire X4270 M2 racks or 13 for all other racks.
Click Save Changes.
Add overrides for nodes 3 and 4 (or nodes 1 and 2 in a six-node cluster). See "Map and Reduce Resource Configuration".
Click the mapreduce1 service to display the mapreduce page:
Expand the Actions menu and select Enable High Availability to start the Enable JobTracker High Availability wizard:
On the Assign Roles page, select the fourth node (node04) for the Standby JobTracker.
Complete the steps of the wizard with no further changes. Click Finish to save the configuration and exit.
Verify that all services in the cluster are healthy with no configuration issues.
Reconfigure Perfect Balance for the MRv1 cluster:
Log in as root to a node of the cluster.
Configure Perfect Balance on all nodes of the cluster:
$ dcli –C /opt/oracle/orabalancer-[version]/bin/configure.sh
See Oracle Big Data Appliance Owner's Guide for the dcli command.
You can take precautions to prevent unauthorized use of the software and data on Oracle Big Data Appliance.
This section contains these topics:
Every open-source package installed on Oracle Big Data Appliance creates one or more users and groups. Most of these users do not have login privileges, shells, or home directories. They are used by daemons and are not intended as an interface for individual users. For example, Hadoop operates as the hdfs user, MapReduce operates as mapred, and Hive operates as hive.
You can use the oracle identity to run Hadoop and Hive jobs immediately after the Oracle Big Data Appliance software is installed. This user account has login privileges, a shell, and a home directory.
Oracle NoSQL Database and Oracle Data Integrator run as the oracle user. Its primary group is oinstall.
Note:
Do not delete, re-create, or modify the users that are created during installation, because they are required for the software to operate.Table 2-9 identifies the operating system users and groups that are created automatically during installation of Oracle Big Data Appliance software for use by CDH components and other software packages.
Table 2-9 Operating System Users and Groups
| User Name | Group | Used By | Login Rights |
|---|---|---|---|
|
|
Apache Flume parent and nodes |
No |
|
|
|
Apache HBase processes |
No |
|
|
|
No |
||
|
|
No |
||
|
|
Hue processes |
No |
|
|
|
ResourceManager, NodeManager, Hive Thrift daemon |
Yes |
|
|
|
|
Yes |
|
|
|
Oozie server |
No |
|
|
Oracle NoSQL Database, Oracle Loader for Hadoop, Oracle Data Integrator, and the Oracle DBA |
Yes |
||
|
|
Puppet parent (puppet nodes run as |
No |
|
|
|
Apache Sqoop metastore |
No |
|
|
Auto Service Request |
No |
||
|
|
ZooKeeper processes |
No |
Oracle Big Data Appliance supports Kerberos security as a software installation option. See Chapter 3 for details about setting up clients and users to access a Kerberos-protected cluster.
The typical authorization model on Hadoop is at the HDFS file level, such that users either have access to all of the data in the file or none. In contrast, Apache Sentry integrates with the Hive and Impala SQL-query engines to provide fine-grained authorization to data and metadata stored in Hadoop.
Oracle Big Data Appliance automatically configures Sentry during software installation, beginning with Mammoth utility version 2.5.
See Also:
Cloudera Manager Help
Managing Clusters with Cloudera Manager at
On-disk encryption protects data that is at rest on disk. When on-disk encryption is enabled, Oracle Big Data Appliance automatically encrypts and decrypts data stored on disk. On-disk encryption does not affect user access to Hadoop data, although it can have a minor impact on performance.
Password-based encryption encodes Hadoop data based on a password, which is the same for all servers in a cluster. You can change the password at any time by using the mammoth-reconfig update command. See Oracle Big Data Appliance Owner's Guide.
If a disk is removed from a server, then the encrypted data remains protected until you install the disk in a server (the same server or a different one), startup the server, and provide the password. If a server is powered off and removed from an Oracle Big Data Appliance rack, then the encrypted data remains protected until you restart server and provide the password. You must enter the password after every startup of every server to enable access to the data. See "Starting Oracle Big Data Appliance".
On-disk encryption is an option that you can select during the initial installation of the software by the Mammoth utility. You can also enable or disable on-disk encryption at any time by using either the mammoth-reconfig or bdacli utilities.
See Also:
Oracle Big Data Appliance Owner's GuideTable 2-10 identifies the port numbers that might be used in addition to those used by CDH.
To view the ports used on a particular server:
In Cloudera Manager, click the Hosts tab at the top of the page to display the Hosts page.
In the Name column, click a server link to see its detail page.
Scroll down to the Ports section.
See Also:
For the full list of CDH port numbers, go to the Cloudera website atThe puppet node service (puppetd) runs continuously as root on all servers. It listens on port 8139 for "kick" requests, which trigger it to request updates from the puppet master. It does not receive updates on this port.
The puppet master service (puppetmasterd) runs continuously as the puppet user on the first server of the primary Oracle Big Data Appliance rack. It listens on port 8140 for requests to push updates to puppet nodes.
The puppet nodes generate and send certificates to the puppet master to register initially during installation of the software. For updates to the software, the puppet master signals ("kicks") the puppet nodes, which then request all configuration changes from the puppet master node that they are registered with.
The puppet master sends updates only to puppet nodes that have known, valid certificates. Puppet nodes only accept updates from the puppet master host name they initially registered with. Because Oracle Big Data Appliance uses an internal network for communication within the rack, the puppet master host name resolves using /etc/hosts to an internal, private IP address.
You can use Oracle Audit Vault and Database Firewall to create and monitor the audit trails for HDFS and MapReduce on Oracle Big Data Appliance.
This section describes the Oracle Big Data Appliance plug-in:
Oracle Audit Vault and Database Firewall secures databases and other critical components of IT infrastructure in these key ways:
Provides an integrated auditing platform for your enterprise.
Captures activity on Oracle Database, Oracle Big Data Appliance, operating systems, directories, file systems, and so forth.
Makes the auditing information available in a single reporting framework so that you can understand the activities across the enterprise. You do not need to monitor each system individually; you can view your computer infrastructure as a whole.
Audit Vault Server provides a web-based, graphic user interface for both administrators and auditors.
You can configure CDH/Hadoop clusters on Oracle Big Data Appliance as secured targets. The Audit Vault plug-in on Oracle Big Data Appliance collects audit and logging data from these services:
Hive DDL: Who makes Hive database changes.
MapReduce: Who runs MapReduce jobs that correspond to file access.
Oozie workflows: Who runs workflow activities.
The Audit Vault plug-in is an installation option. The Mammoth utility automatically configures monitoring on Oracle Big Data Appliance as part of the software installation process.
See Also:
For more information about Oracle Audit Vault and Database Firewall:The Mammoth utility on Oracle Big Data Appliance performs all the steps needed to setup the plug-in, using information that you provide.
To set up the Audit Vault plug-in for Oracle Big Data Appliance:
Ensure that Oracle Audit Vault and Database Firewall Server Release 12.1.1 is up and running on the same network as Oracle Big Data Appliance.
Complete the Audit Vault Plug-in section of Oracle Big Data Appliance Configuration Generation Utility.
Install the Oracle Big Data Appliance software using the Mammoth utility. An Oracle representative typically performs this step.
You can also add the plug-in at a later time using either bdacli or mammoth-reconfig. See Oracle Big Data Appliance Owner's Guide.
When the software installation is complete, the Audit Vault plug-in is installed on Oracle Big Data Appliance, and Oracle Audit Vault and Database Firewall is collecting its audit information. You do not need to perform any other installation steps.
See Also:
Oracle Big Data Appliance Owner's Guide for using Oracle Big Data Appliance Configuration Generation UtilityAfter installing the plug-in, you can monitor Oracle Big Data Appliance the same as any other secured target. Audit Vault Server collects activity reports automatically.
The following procedure describes one type of monitoring activity.
To view an Oracle Big Data Appliance activity report:
Log in to Audit Vault Server as an auditor.
Click the Reports tab.
Under Built-in Reports, click Audit Reports.
To browse all activities, in the Activity Reports list, click the Browse report data icon for All Activity.
Add or remove the filters to list the events.
Event names include ACCESS, CREATE, DELETE, and OPEN.
Click the Single row view icon in the first column to see a detailed report.
Figure 2-10 shows the beginning of an activity report, which records access to a Hadoop sequence file.
Figure 2-10 Activity Report in Audit Vault Server

If you need help from Oracle Support to troubleshoot CDH issues, then you should first collect diagnostic information using the bdadiag utility with the cm option.
To collect diagnostic information:
Log in to an Oracle Big Data Appliance server as root.
Run bdadiag with at least the cm option. You can include additional options on the command line as appropriate. See the Oracle Big Data Appliance Owner's Guide for a complete description of the bdadiag syntax.
# bdadiag cm
The command output identifies the name and the location of the diagnostic file.
Go to My Oracle Support at http://support.oracle.com.
Open a Service Request (SR) if you have not already done so.
Upload the bz2 file into the SR. If the file is too large, then upload it to sftp.oracle.com, as described in the next procedure.
To upload the diagnostics to ftp.oracle.com:
Open an SFTP client and connect to sftp.oracle.com. Specify port 2021 and remote directory /support/incoming/target, where target is the folder name given to you by Oracle Support.
See Example 2-1 if you are using a command-line SFTP client.
Log in with your Oracle Single Sign-on account and password.
Upload the diagnostic file to the new directory.
Update the SR with the full path and the file name.
Example 2-1 shows the commands to upload the diagnostics using the SFTP command interface.
Example 2-1 Uploading Diagnostics Using FTP
$ sftp -o port=2021 my.user.name@oracle.com@sftp.oracle.com Connecting to sftp.oracle.com... . . . Enter password for my.user.name@oracle.com Password: password sftp> cd support/incoming/SR123456 sftp> put /tmp/bdadiag_bda1node01_1216FM5497_2013_07_18_07_33.tar.bz2 Uploading bdadiag_bda1node01_1216FM5497_2013_07_18_07_33.tar.bz2 to //support/incoming/create_table.sql bdadiag_bda1node01_1216FM5497_2013_07_18_07_33.tar.bz2 to support/incoming/create_table.sql 100% 311 0.3KB/s 00:00 sftp> exit $
See Also:
My Oracle Support Note 549180.1 at